Dynamic Regret of Online Mirror Descent for Relatively Smooth Convex Cost Functions
نویسندگان
چکیده
The performance of online convex optimization algorithms in a dynamic environment is often expressed terms the regret, which measures decision maker’s against sequence time-varying comparators. In analysis prior works assume Lipschitz continuity or uniform smoothness cost functions. However, there are many important functions practice that do not satisfy these conditions. such cases, analyses applicable and fail to guarantee performance. this letter, we show it possible bound even when neither nor present. We adopt notion relative with respect some user-defined regularization function, much milder requirement on first under smoothness, regret has an upper based path length functional variation. then additional condition relatively strong convexity, can be bounded by gradient These bounds provide guarantees wide variety problems arise different application domains. Finally, present numerical experiments demonstrate advantage adopting function smooth.
منابع مشابه
Shifting Regret, Mirror Descent, and Matrices
We consider the problem of online prediction in changing environments. In this framework the performance of a predictor is evaluated as the loss relative to an arbitrarily changing predictor, whose individual components come from a base class of predictors. Typical results in the literature consider different base classes (experts, linear predictors on the simplex, etc.) separately. Introducing...
متن کاملNo-regret Algorithms for Online Convex Programs
Online convex programming has recently emerged as a powerful primitive for designing machine learning algorithms. For example, OCP can be used for learning a linear classifier, dynamically rebalancing a binary search tree, finding the shortest path in a graph with unknown edge lengths, solving a structured classification problem, or finding a good strategy in an extensive-form game. Several res...
متن کاملNo-regret algorithms for Online Convex Programs
Online convex programming has recently emerged as a powerful primitive for designing machine learning algorithms. For example, OCP can be used for learning a linear classifier, dynamically rebalancing a binary search tree, finding the shortest path in a graph with unknown edge lengths, solving a structured classification problem, or finding a good strategy in an extensive-form game. Several res...
متن کاملMirror descent in non-convex stochastic programming
In this paper, we examine a class of nonconvex stochastic optimization problems which we call variationally coherent, and which properly includes all quasi-convex programs. In view of solving such problems, we focus on the widely used stochastic mirror descent (SMD) family of algorithms, and we establish that the method’s last iterate converges with probability 1. We further introduce a localiz...
متن کاملConvergence of Online Mirror Descent Algorithms
In this paper we consider online mirror descent (OMD) algorithms, a class of scalable online learning algorithms exploiting data geometric structures through mirror maps. Necessary and sufficient conditions are presented in terms of the step size sequence {ηt}t for the convergence of an OMD algorithm with respect to the expected Bregman distance induced by the mirror map. The condition is limt→...
متن کاملذخیره در منابع من
با ذخیره ی این منبع در منابع من، دسترسی به آن را برای استفاده های بعدی آسان تر کنید
ژورنال
عنوان ژورنال: IEEE Control Systems Letters
سال: 2022
ISSN: ['2475-1456']
DOI: https://doi.org/10.1109/lcsys.2022.3155067